# Multi-stage fine-tuning

Acemath 72B Instruct
AceMath is a series of cutting-edge models designed specifically for mathematical reasoning. It is improved based on Qwen and is good at using chain-of-thought (CoT) reasoning to solve English mathematical problems.
Large Language Model Safetensors English
A
nvidia
3,141
18
AMD OLMo 1B
Apache-2.0
AMD-OLMo is a series of 1-billion-parameter language models trained from scratch by AMD on the AMD Instinct™ MI250 GPU.
Large Language Model Safetensors
A
amd
4,419
25
Trocr Medieval Castilian Hybrida
MIT
This is a TrOCR model specifically fine-tuned for medieval Castilian (particularly Hybrida handwriting), based on microsoft/trocr-base-handwritten.
Text Recognition Transformers Other
T
medieval-data
33
1
Llama 3 8B Instruct Gradient 4194k
An extended long-context model based on Meta-Llama-3-8B-Instruct, achieving 4194K context length support through RoPE theta parameter adjustment
Large Language Model Transformers English
L
gradientai
244
70
Sinhala Ocr Model V3
This model is a fine-tuned Sinhala OCR model based on Ransaka/sinhala-ocr-model, primarily used for recognizing Sinhala text.
Text Recognition Transformers
S
Ransaka
213
2
Beit Large Patch16 224
Apache-2.0
BEiT is an image classification model based on Vision Transformer (ViT) architecture, pretrained with self-supervised learning on ImageNet-21k and fine-tuned on ImageNet-1k.
Image Classification
B
microsoft
222.46k
1
En Vi
Apache-2.0
An English-Vietnamese translation model fine-tuned based on the MarianMT pre-trained model, employing a two-phase training strategy
Machine Translation Transformers Supports Multiple Languages
E
CLAck
31
1
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase